-
1 formal reasoning
формальные рассуждения; рассуждения на формальном уровне -
2 formal reasoning
Большой англо-русский и русско-английский словарь > formal reasoning
-
3 formal reasoning
Англо-русский словарь технических терминов > formal reasoning
-
4 formal reasoning
1) Техника: формальное рассуждение2) Вычислительная техника: рассуждения на формальном уровне, формальные рассуждения -
5 formal reasoning
формальные рассуждения; рассуждения на формальном уровнеEnglish-Russian dictionary of computer science and programming > formal reasoning
-
6 formal reasoning
-
7 formal reasoning
Англо-русский словарь по исследованиям и ноу-хау > formal reasoning
-
8 reasoning
1) рассуждение; рассуждения2) мышление3) обоснование•- bottom-up reasoning
- case-based reasoning
- commonsense reasoning
- deductive reasoning
- default reasoning
- defeasible reasoning
- enumerative reasoning
- evidential reasoning
- exact reasoning
- formal reasoning
- hypothetical reasoning
- inexact reasoning
- integrating reasoning
- meta-level reasoning
- model-based reasoning
- plausible reasoning
- procedural reasoning
- reasoning by analogy
- reflective reasoning
- spatial reasoning
- taxonomic reasoningEnglish-Russian dictionary of computer science and programming > reasoning
-
9 reasoning
2) рассуждение• -
10 reasoning
(логическое) рассуждение; (логический) вывод -
11 formal
1 ( official) [agreement, announcement, application, complaint, enquiry, interview, invitation, protest, reception] officiel/-ielle ;2 ( not casual) [language, register, style] soutenu ; [occasion] solennel/-elle ; [welcome, manner] cérémonieux/-ieuse ; [clothing, outfit, jacket] habillé ; ( on invitation) ‘dress: formal’ ‘tenue de soirée’ ; ‘assistance’ is a formal word for ‘help’ ‘assistance’ est plus soutenu que ‘aide’ ; he sounded very formal il avait l'air très guindé pej ; formal teaching methods méthodes traditionnelles d'enseignement ;3 ( structured) [logic, proof, grammar, linguistics, reasoning] formel/-elle ;4 ( in recognized institution) [training] professionnel/-elle ; [qualification] reconnu ; he had no formal education il n'était jamais allé à l'école ; -
12 Reasoning
For some considerable time we cherished the illusion that [using formal logic to construct psychological models of reasoning] was the way to proceed and that only the structural characteristics of the problem mattered. Only gradually did we realise first that there was no existing formal calculus which correctly modelled our subject's inferences, and second that no purely formal calculus would succeed. (Wason & Johnson-Laird, 1972, p. 244)Historical dictionary of quotations in cognitive science > Reasoning
-
13 formal system of reasoning
Лингвистика: формальная система рассужденийУниверсальный англо-русский словарь > formal system of reasoning
-
14 формальные рассуждения
Большой англо-русский и русско-английский словарь > формальные рассуждения
-
15 Artificial Intelligence
In my opinion, none of [these programs] does even remote justice to the complexity of human mental processes. Unlike men, "artificially intelligent" programs tend to be single minded, undistractable, and unemotional. (Neisser, 1967, p. 9)Future progress in [artificial intelligence] will depend on the development of both practical and theoretical knowledge.... As regards theoretical knowledge, some have sought a unified theory of artificial intelligence. My view is that artificial intelligence is (or soon will be) an engineering discipline since its primary goal is to build things. (Nilsson, 1971, pp. vii-viii)Most workers in AI [artificial intelligence] research and in related fields confess to a pronounced feeling of disappointment in what has been achieved in the last 25 years. Workers entered the field around 1950, and even around 1960, with high hopes that are very far from being realized in 1972. In no part of the field have the discoveries made so far produced the major impact that was then promised.... In the meantime, claims and predictions regarding the potential results of AI research had been publicized which went even farther than the expectations of the majority of workers in the field, whose embarrassments have been added to by the lamentable failure of such inflated predictions....When able and respected scientists write in letters to the present author that AI, the major goal of computing science, represents "another step in the general process of evolution"; that possibilities in the 1980s include an all-purpose intelligence on a human-scale knowledge base; that awe-inspiring possibilities suggest themselves based on machine intelligence exceeding human intelligence by the year 2000 [one has the right to be skeptical]. (Lighthill, 1972, p. 17)4) Just as Astronomy Succeeded Astrology, the Discovery of Intellectual Processes in Machines Should Lead to a Science, EventuallyJust as astronomy succeeded astrology, following Kepler's discovery of planetary regularities, the discoveries of these many principles in empirical explorations on intellectual processes in machines should lead to a science, eventually. (Minsky & Papert, 1973, p. 11)5) Problems in Machine Intelligence Arise Because Things Obvious to Any Person Are Not Represented in the ProgramMany problems arise in experiments on machine intelligence because things obvious to any person are not represented in any program. One can pull with a string, but one cannot push with one.... Simple facts like these caused serious problems when Charniak attempted to extend Bobrow's "Student" program to more realistic applications, and they have not been faced up to until now. (Minsky & Papert, 1973, p. 77)What do we mean by [a symbolic] "description"? We do not mean to suggest that our descriptions must be made of strings of ordinary language words (although they might be). The simplest kind of description is a structure in which some features of a situation are represented by single ("primitive") symbols, and relations between those features are represented by other symbols-or by other features of the way the description is put together. (Minsky & Papert, 1973, p. 11)[AI is] the use of computer programs and programming techniques to cast light on the principles of intelligence in general and human thought in particular. (Boden, 1977, p. 5)The word you look for and hardly ever see in the early AI literature is the word knowledge. They didn't believe you have to know anything, you could always rework it all.... In fact 1967 is the turning point in my mind when there was enough feeling that the old ideas of general principles had to go.... I came up with an argument for what I called the primacy of expertise, and at the time I called the other guys the generalists. (Moses, quoted in McCorduck, 1979, pp. 228-229)9) Artificial Intelligence Is Psychology in a Particularly Pure and Abstract FormThe basic idea of cognitive science is that intelligent beings are semantic engines-in other words, automatic formal systems with interpretations under which they consistently make sense. We can now see why this includes psychology and artificial intelligence on a more or less equal footing: people and intelligent computers (if and when there are any) turn out to be merely different manifestations of the same underlying phenomenon. Moreover, with universal hardware, any semantic engine can in principle be formally imitated by a computer if only the right program can be found. And that will guarantee semantic imitation as well, since (given the appropriate formal behavior) the semantics is "taking care of itself" anyway. Thus we also see why, from this perspective, artificial intelligence can be regarded as psychology in a particularly pure and abstract form. The same fundamental structures are under investigation, but in AI, all the relevant parameters are under direct experimental control (in the programming), without any messy physiology or ethics to get in the way. (Haugeland, 1981b, p. 31)There are many different kinds of reasoning one might imagine:Formal reasoning involves the syntactic manipulation of data structures to deduce new ones following prespecified rules of inference. Mathematical logic is the archetypical formal representation. Procedural reasoning uses simulation to answer questions and solve problems. When we use a program to answer What is the sum of 3 and 4? it uses, or "runs," a procedural model of arithmetic. Reasoning by analogy seems to be a very natural mode of thought for humans but, so far, difficult to accomplish in AI programs. The idea is that when you ask the question Can robins fly? the system might reason that "robins are like sparrows, and I know that sparrows can fly, so robins probably can fly."Generalization and abstraction are also natural reasoning process for humans that are difficult to pin down well enough to implement in a program. If one knows that Robins have wings, that Sparrows have wings, and that Blue jays have wings, eventually one will believe that All birds have wings. This capability may be at the core of most human learning, but it has not yet become a useful technique in AI.... Meta- level reasoning is demonstrated by the way one answers the question What is Paul Newman's telephone number? You might reason that "if I knew Paul Newman's number, I would know that I knew it, because it is a notable fact." This involves using "knowledge about what you know," in particular, about the extent of your knowledge and about the importance of certain facts. Recent research in psychology and AI indicates that meta-level reasoning may play a central role in human cognitive processing. (Barr & Feigenbaum, 1981, pp. 146-147)Suffice it to say that programs already exist that can do things-or, at the very least, appear to be beginning to do things-which ill-informed critics have asserted a priori to be impossible. Examples include: perceiving in a holistic as opposed to an atomistic way; using language creatively; translating sensibly from one language to another by way of a language-neutral semantic representation; planning acts in a broad and sketchy fashion, the details being decided only in execution; distinguishing between different species of emotional reaction according to the psychological context of the subject. (Boden, 1981, p. 33)Can the synthesis of Man and Machine ever be stable, or will the purely organic component become such a hindrance that it has to be discarded? If this eventually happens-and I have... good reasons for thinking that it must-we have nothing to regret and certainly nothing to fear. (Clarke, 1984, p. 243)The thesis of GOFAI... is not that the processes underlying intelligence can be described symbolically... but that they are symbolic. (Haugeland, 1985, p. 113)14) Artificial Intelligence Provides a Useful Approach to Psychological and Psychiatric Theory FormationIt is all very well formulating psychological and psychiatric theories verbally but, when using natural language (even technical jargon), it is difficult to recognise when a theory is complete; oversights are all too easily made, gaps too readily left. This is a point which is generally recognised to be true and it is for precisely this reason that the behavioural sciences attempt to follow the natural sciences in using "classical" mathematics as a more rigorous descriptive language. However, it is an unfortunate fact that, with a few notable exceptions, there has been a marked lack of success in this application. It is my belief that a different approach-a different mathematics-is needed, and that AI provides just this approach. (Hand, quoted in Hand, 1985, pp. 6-7)We might distinguish among four kinds of AI.Research of this kind involves building and programming computers to perform tasks which, to paraphrase Marvin Minsky, would require intelligence if they were done by us. Researchers in nonpsychological AI make no claims whatsoever about the psychological realism of their programs or the devices they build, that is, about whether or not computers perform tasks as humans do.Research here is guided by the view that the computer is a useful tool in the study of mind. In particular, we can write computer programs or build devices that simulate alleged psychological processes in humans and then test our predictions about how the alleged processes work. We can weave these programs and devices together with other programs and devices that simulate different alleged mental processes and thereby test the degree to which the AI system as a whole simulates human mentality. According to weak psychological AI, working with computer models is a way of refining and testing hypotheses about processes that are allegedly realized in human minds.... According to this view, our minds are computers and therefore can be duplicated by other computers. Sherry Turkle writes that the "real ambition is of mythic proportions, making a general purpose intelligence, a mind." (Turkle, 1984, p. 240) The authors of a major text announce that "the ultimate goal of AI research is to build a person or, more humbly, an animal." (Charniak & McDermott, 1985, p. 7)Research in this field, like strong psychological AI, takes seriously the functionalist view that mentality can be realized in many different types of physical devices. Suprapsychological AI, however, accuses strong psychological AI of being chauvinisticof being only interested in human intelligence! Suprapsychological AI claims to be interested in all the conceivable ways intelligence can be realized. (Flanagan, 1991, pp. 241-242)16) Determination of Relevance of Rules in Particular ContextsEven if the [rules] were stored in a context-free form the computer still couldn't use them. To do that the computer requires rules enabling it to draw on just those [ rules] which are relevant in each particular context. Determination of relevance will have to be based on further facts and rules, but the question will again arise as to which facts and rules are relevant for making each particular determination. One could always invoke further facts and rules to answer this question, but of course these must be only the relevant ones. And so it goes. It seems that AI workers will never be able to get started here unless they can settle the problem of relevance beforehand by cataloguing types of context and listing just those facts which are relevant in each. (Dreyfus & Dreyfus, 1986, p. 80)Perhaps the single most important idea to artificial intelligence is that there is no fundamental difference between form and content, that meaning can be captured in a set of symbols such as a semantic net. (G. Johnson, 1986, p. 250)Artificial intelligence is based on the assumption that the mind can be described as some kind of formal system manipulating symbols that stand for things in the world. Thus it doesn't matter what the brain is made of, or what it uses for tokens in the great game of thinking. Using an equivalent set of tokens and rules, we can do thinking with a digital computer, just as we can play chess using cups, salt and pepper shakers, knives, forks, and spoons. Using the right software, one system (the mind) can be mapped into the other (the computer). (G. Johnson, 1986, p. 250)19) A Statement of the Primary and Secondary Purposes of Artificial IntelligenceThe primary goal of Artificial Intelligence is to make machines smarter.The secondary goals of Artificial Intelligence are to understand what intelligence is (the Nobel laureate purpose) and to make machines more useful (the entrepreneurial purpose). (Winston, 1987, p. 1)The theoretical ideas of older branches of engineering are captured in the language of mathematics. We contend that mathematical logic provides the basis for theory in AI. Although many computer scientists already count logic as fundamental to computer science in general, we put forward an even stronger form of the logic-is-important argument....AI deals mainly with the problem of representing and using declarative (as opposed to procedural) knowledge. Declarative knowledge is the kind that is expressed as sentences, and AI needs a language in which to state these sentences. Because the languages in which this knowledge usually is originally captured (natural languages such as English) are not suitable for computer representations, some other language with the appropriate properties must be used. It turns out, we think, that the appropriate properties include at least those that have been uppermost in the minds of logicians in their development of logical languages such as the predicate calculus. Thus, we think that any language for expressing knowledge in AI systems must be at least as expressive as the first-order predicate calculus. (Genesereth & Nilsson, 1987, p. viii)21) Perceptual Structures Can Be Represented as Lists of Elementary PropositionsIn artificial intelligence studies, perceptual structures are represented as assemblages of description lists, the elementary components of which are propositions asserting that certain relations hold among elements. (Chase & Simon, 1988, p. 490)Artificial intelligence (AI) is sometimes defined as the study of how to build and/or program computers to enable them to do the sorts of things that minds can do. Some of these things are commonly regarded as requiring intelligence: offering a medical diagnosis and/or prescription, giving legal or scientific advice, proving theorems in logic or mathematics. Others are not, because they can be done by all normal adults irrespective of educational background (and sometimes by non-human animals too), and typically involve no conscious control: seeing things in sunlight and shadows, finding a path through cluttered terrain, fitting pegs into holes, speaking one's own native tongue, and using one's common sense. Because it covers AI research dealing with both these classes of mental capacity, this definition is preferable to one describing AI as making computers do "things that would require intelligence if done by people." However, it presupposes that computers could do what minds can do, that they might really diagnose, advise, infer, and understand. One could avoid this problematic assumption (and also side-step questions about whether computers do things in the same way as we do) by defining AI instead as "the development of computers whose observable performance has features which in humans we would attribute to mental processes." This bland characterization would be acceptable to some AI workers, especially amongst those focusing on the production of technological tools for commercial purposes. But many others would favour a more controversial definition, seeing AI as the science of intelligence in general-or, more accurately, as the intellectual core of cognitive science. As such, its goal is to provide a systematic theory that can explain (and perhaps enable us to replicate) both the general categories of intentionality and the diverse psychological capacities grounded in them. (Boden, 1990b, pp. 1-2)Because the ability to store data somewhat corresponds to what we call memory in human beings, and because the ability to follow logical procedures somewhat corresponds to what we call reasoning in human beings, many members of the cult have concluded that what computers do somewhat corresponds to what we call thinking. It is no great difficulty to persuade the general public of that conclusion since computers process data very fast in small spaces well below the level of visibility; they do not look like other machines when they are at work. They seem to be running along as smoothly and silently as the brain does when it remembers and reasons and thinks. On the other hand, those who design and build computers know exactly how the machines are working down in the hidden depths of their semiconductors. Computers can be taken apart, scrutinized, and put back together. Their activities can be tracked, analyzed, measured, and thus clearly understood-which is far from possible with the brain. This gives rise to the tempting assumption on the part of the builders and designers that computers can tell us something about brains, indeed, that the computer can serve as a model of the mind, which then comes to be seen as some manner of information processing machine, and possibly not as good at the job as the machine. (Roszak, 1994, pp. xiv-xv)The inner workings of the human mind are far more intricate than the most complicated systems of modern technology. Researchers in the field of artificial intelligence have been attempting to develop programs that will enable computers to display intelligent behavior. Although this field has been an active one for more than thirty-five years and has had many notable successes, AI researchers still do not know how to create a program that matches human intelligence. No existing program can recall facts, solve problems, reason, learn, and process language with human facility. This lack of success has occurred not because computers are inferior to human brains but rather because we do not yet know in sufficient detail how intelligence is organized in the brain. (Anderson, 1995, p. 2)Historical dictionary of quotations in cognitive science > Artificial Intelligence
-
16 формальное рассуждение
Большой англо-русский и русско-английский словарь > формальное рассуждение
-
17 Thinking
But what then am I? A thing which thinks. What is a thing which thinks? It is a thing which doubts, understands, [conceives], affirms, denies, wills, refuses, which also imagines and feels. (Descartes, 1951, p. 153)I have been trying in all this to remove the temptation to think that there "must be" a mental process of thinking, hoping, wishing, believing, etc., independent of the process of expressing a thought, a hope, a wish, etc.... If we scrutinize the usages which we make of "thinking," "meaning," "wishing," etc., going through this process rids us of the temptation to look for a peculiar act of thinking, independent of the act of expressing our thoughts, and stowed away in some particular medium. (Wittgenstein, 1958, pp. 41-43)Analyse the proofs employed by the subject. If they do not go beyond observation of empirical correspondences, they can be fully explained in terms of concrete operations, and nothing would warrant our assuming that more complex thought mechanisms are operating. If, on the other hand, the subject interprets a given correspondence as the result of any one of several possible combinations, and this leads him to verify his hypotheses by observing their consequences, we know that propositional operations are involved. (Inhelder & Piaget, 1958, p. 279)In every age, philosophical thinking exploits some dominant concepts and makes its greatest headway in solving problems conceived in terms of them. The seventeenth- and eighteenth-century philosophers construed knowledge, knower, and known in terms of sense data and their association. Descartes' self-examination gave classical psychology the mind and its contents as a starting point. Locke set up sensory immediacy as the new criterion of the real... Hobbes provided the genetic method of building up complex ideas from simple ones... and, in another quarter, still true to the Hobbesian method, Pavlov built intellect out of conditioned reflexes and Loeb built life out of tropisms. (S. Langer, 1962, p. 54)Experiments on deductive reasoning show that subjects are influenced sufficiently by their experience for their reasoning to differ from that described by a purely deductive system, whilst experiments on inductive reasoning lead to the view that an understanding of the strategies used by adult subjects in attaining concepts involves reference to higher-order concepts of a logical and deductive nature. (Bolton, 1972, p. 154)There are now machines in the world that think, that learn and create. Moreover, their ability to do these things is going to increase rapidly until-in the visible future-the range of problems they can handle will be coextensive with the range to which the human mind has been applied. (Newell & Simon, quoted in Weizenbaum, 1976, p. 138)But how does it happen that thinking is sometimes accompanied by action and sometimes not, sometimes by motion, and sometimes not? It looks as if almost the same thing happens as in the case of reasoning and making inferences about unchanging objects. But in that case the end is a speculative proposition... whereas here the conclusion which results from the two premises is an action.... I need covering; a cloak is a covering. I need a cloak. What I need, I have to make; I need a cloak. I have to make a cloak. And the conclusion, the "I have to make a cloak," is an action. (Nussbaum, 1978, p. 40)It is well to remember that when philosophy emerged in Greece in the sixth century, B.C., it did not burst suddenly out of the Mediterranean blue. The development of societies of reasoning creatures-what we call civilization-had been a process to be measured not in thousands but in millions of years. Human beings became civilized as they became reasonable, and for an animal to begin to reason and to learn how to improve its reasoning is a long, slow process. So thinking had been going on for ages before Greece-slowly improving itself, uncovering the pitfalls to be avoided by forethought, endeavoring to weigh alternative sets of consequences intellectually. What happened in the sixth century, B.C., is that thinking turned round on itself; people began to think about thinking, and the momentous event, the culmination of the long process to that point, was in fact the birth of philosophy. (Lipman, Sharp & Oscanyan, 1980, p. xi)The way to look at thought is not to assume that there is a parallel thread of correlated affects or internal experiences that go with it in some regular way. It's not of course that people don't have internal experiences, of course they do; but that when you ask what is the state of mind of someone, say while he or she is performing a ritual, it's hard to believe that such experiences are the same for all people involved.... The thinking, and indeed the feeling in an odd sort of way, is really going on in public. They are really saying what they're saying, doing what they're doing, meaning what they're meaning. Thought is, in great part anyway, a public activity. (Geertz, quoted in J. Miller, 1983, pp. 202-203)Everything should be made as simple as possible, but not simpler. (Einstein, quoted in Minsky, 1986, p. 17)What, in effect, are the conditions for the construction of formal thought? The child must not only apply operations to objects-in other words, mentally execute possible actions on them-he must also "reflect" those operations in the absence of the objects which are replaced by pure propositions. Thus, "reflection" is thought raised to the second power. Concrete thinking is the representation of a possible action, and formal thinking is the representation of a representation of possible action.... It is not surprising, therefore, that the system of concrete operations must be completed during the last years of childhood before it can be "reflected" by formal operations. In terms of their function, formal operations do not differ from concrete operations except that they are applied to hypotheses or propositions [whose logic is] an abstract translation of the system of "inference" that governs concrete operations. (Piaget, quoted in Minsky, 1986, p. 237)[E]ven a human being today (hence, a fortiori, a remote ancestor of contemporary human beings) cannot easily or ordinarily maintain uninterrupted attention on a single problem for more than a few tens of seconds. Yet we work on problems that require vastly more time. The way we do that (as we can observe by watching ourselves) requires periods of mulling to be followed by periods of recapitulation, describing to ourselves what seems to have gone on during the mulling, leading to whatever intermediate results we have reached. This has an obvious function: namely, by rehearsing these interim results... we commit them to memory, for the immediate contents of the stream of consciousness are very quickly lost unless rehearsed.... Given language, we can describe to ourselves what seemed to occur during the mulling that led to a judgment, produce a rehearsable version of the reaching-a-judgment process, and commit that to long-term memory by in fact rehearsing it. (Margolis, 1987, p. 60)Historical dictionary of quotations in cognitive science > Thinking
-
18 sound
I
adjective1) (strong or in good condition: The foundations of the house are not very sound; He's 87, but he's still sound in mind and body.) sano, sólido, firme2) ((of sleep) deep: She's a very sound sleeper.) profundo3) (full; thorough: a sound basic training.) completo, severo4) (accurate; free from mistakes: a sound piece of work.) bueno, sólido5) (having or showing good judgement or good sense: His advice is always very sound.) juicioso, sensato, acertado, consistente•- soundly- soundness
- sound asleep
II
1. noun1) (the impressions transmitted to the brain by the sense of hearing: a barrage of sound; (also adjective) sound waves.) sonido2) (something that is, or can be, heard: The sounds were coming from the garage.) ruido, sonido, voces3) (the impression created in the mind by a piece of news, a description etc: I didn't like the sound of her hairstyle at all!) idea
2. verb1) (to (cause something to) make a sound: Sound the bell!; The bell sounded.) tocar, hacer sonar2) (to signal (something) by making a sound: Sound the alarm!) sonar, resonar3) ((of something heard or read) to make a particular impression; to seem; to appear: Your singing sounded very good; That sounds like a train.) sonar, parecer4) (to pronounce: In the word `pneumonia', the letter p is not sounded.) pronunciar(se)5) (to examine by tapping and listening carefully: She sounded the patient's chest.) auscultar•- soundlessly
- sound effects
- soundproof
3. verb(to make (walls, a room etc) soundproof.) insonorizar
III
verb(to measure the depth of (water etc).) sondar- sounding- sound out
sound1 adj sólido / bueno / fuerte / segurosound2 n1. sonido2. ruido3. volumencan you turn the sound up? ¿puedes subir el volumen?sound3 vb1. parecerfrom what he says it sounds like a wonderful hotel por lo que dice, parece un hotel maravilloso2. sonarif the alarm sounds, leave the building si suena la alarma, abandona el edificiotr[saʊnd]1 (healthy) sano,-a■ of sound mind en su sano juicio, en pleno uso de sus facultades■ safe and sound sano,-a y salvo,-a2 (solid) sólido,-a, firme; (in good condition) en buen estado3 (sensible) sensato,-a, acertado,-a; (valid) sólido,-a, lógico,-a, razonable; (responsible) responsable, formal, de fiar; (reliable, safe) seguro,-a4 (thorough) completo,-a; (severe) severo,-a5 (of sleep) profundo,-a\SMALLIDIOMATIC EXPRESSION/SMALLto be as sound as a bell (person) estar sano,-a 2 (thing) estar en perfectas condiciones, estar en perfecto estadoto be sound asleep estar profundamente dormido,-a————————tr[saʊnd]1 SMALLGEOGRAPHY/SMALL estrecho, brazo de mar————————tr[saʊnd]■ I was born within the sound of Bow bells desde donde nací se oyen las campanas de la iglesia de Bow■ turn the sound up/down sube/baja el volumen3 (impression, idea) idea■ I don't like the sound of this esto se está poniendo feo, esto me da mala espina■ by/from the sound of it, he's getting on fine por lo visto las cosas le van bien2 SMALLLINGUISTICS/SMALL pronunciar1 (bell, horn, alarm, etc) sonar, resonar2 (seem) parecer; (give impression) sonar■ how does that sound? ¿qué te parece eso?■ does this sentence sound right to you? ¿te suena bien esta frase?3 SMALLLINGUISTICS/SMALL pronunciarse, sonar\SMALLIDIOMATIC EXPRESSION/SMALLsound barrier barrera del sonidosound card tarjeta de sonidosound check prueba de sonidosound effects efectos nombre masculino plural sonorossound engineer ingeniero,-a de sonidosound wave onda sonora————————tr[saʊnd]1 SMALLMARITIME/SMALL sondar1 SMALLMEDICINE/SMALL sondasound ['saʊnd] vt1) : sondar (en navegación)3) : hacer sonar, tocar (una trompeta, etc.)sound vi1) : sonarthe alarm sounded: la alarma sonó2) seem: parecersound adj1) healthy: sanosafe and sound: sano y salvoof sound mind and body: en pleno uso de sus facultades2) firm, solid: sólido3) sensible: lógico, sensato4) deep: profundoa sound sleep: un sueño profundosound n1) : sonido mthe speed of sound: la velocidad del sonido2) noise: sonido m, ruido mI heard a sound: oí un sonido3) channel: brazo m de mar, canal m (ancho)adj.• confiable adj.• firme adj.• ileso, -a adj.• macizo, -a adj.• razonable adj.• sano, -a adj.• sonido, -a adj.• sólido, -a adj.n.• ruido s.m.• son s.m.• sonda s.f.• sonido s.m.• tañido s.m.• toque s.m.v.• fondear v.• hondear v.• sonar v.• sondear v.• tocar v.
I saʊnd1) noun2) u ca) ( noise) sonido m; (unpleasant, disturbing) ruido mb) (of music, instrument) sonido mc) ( Ling) sonido m3) ua) ( Phys) sonido m; (before n)b) (Audio, Rad, TV) sonido mturn the sound up/down — sube/baja el volumen; (before n)
sound effects — efectos mpl sonoros
4) ( impression conveyed) (colloq) (no pl)by o from the sound of it, everything's going very well — parece que or por lo visto todo marcha muy bien
5) ca) ( channel) paso m, estrecho mb) ( inlet) brazo m
II
1.
1)a) ( give impression) sonar*your voice sounds o you sound different on the phone — tu voz suena distinta por teléfono
you sound as if o as though you could do with a rest — me da la impresión de que no te vendría mal un descanso
it sounds as if o as though they're here now — (por el ruido) parece que ya están aquí
b) ( seem) parecer*we'll leave at ten; how does that sound to you? — saldremos a las diez ¿qué te parece?
it sounds as if o as though you had a great time — parece que lo pasaste fenomenal
sounds like fun! — (colloq) qué divertido!
2) (make noise, resound) \<\<bell/alarm\>\> sonar*
2.
vt1)a) \<\<trumpet/horn\>\> tocar*, hacer* sonarthe chairman sounded a note of warning in his speech — en su discurso, el presidente llamó a la cautela
b) ( articulate) \<\<letter/consonant\>\> pronunciar2) sound out•Phrasal Verbs:
III
adjective -er, -est1)a) ( healthy) sanoI, Peter Smith, being of sound mind... — (frml) yo, Peter Smith, (estando) en pleno uso de mis facultades... (frml)
b) ( in good condition) <basis/foundation> sólido, firme; < timber> en buenas condiciones2)a) ( valid) <reasoning/knowledge> sólido; <advice/decision> sensatob) ( reliable) <colleague/staff> responsable, formal3)b) (hard, thorough)
IV
adverb -er, -est
I [saʊnd]1. N1) (Phys) sonido m2) (=noise) ruido mthe sound of breaking glass — el ruido de cristales que se rompen/rompían
•
I didn't hear a sound — no oí ni un ruido•
don't make a sound! — ¡no hagas el menor ruido!•
not a sound was to be heard — no se oía or (esp LAm) sentía ruido alguno•
to the sound of the national anthem — al son del himno nacional•
they were within sound of the camp — el campamento estaba al alcance del oído•
he opened the door without a sound — abrió la puerta sin hacer nada de ruido3) (=volume) volumen mcan I turn the sound down? — ¿puedo bajar el volumen?
4) (=musical style)5) (fig) (=impression)•
by the sound of it — según parece•
I don't like the sound of it — (film etc) por lo que he oído, no me gusta nada; (situation) me preocupa, me da mala espina2. VT1) [+ horn, trumpet] tocar, hacer sonar; [+ bell] tocar; [+ alarm, warning] dar; [+ praises] cantar, entonar•
to sound the charge — (Mil) tocar la carga•
sound your horn! — (Aut) ¡toca el claxon!•
to sound a note of warning — (fig) dar la señal de alarma•
to sound the retreat — (Mil) tocar la retirada2) (=pronounce) pronunciarsound your "r"s more — pronuncia más claro la "r"
to sound the "d" in "hablado" — pronunciar la "d" en "hablado"
3. VI1) (=emit sound) sonara cannon sounded a long way off — se oyó un cañón a lo lejos, sonó or resonó un cañón a lo lejos
2) (=appear to be)a) (from aural clues) sonarhe sounds Italian to me — por la voz, diría que es italiano
•
it sounds like French — suena a francésb) (from available information) sonar, parecer•
it sounds as if or as though she won't be coming — parece que no va a venir•
how does it sound to you? — ¿qué te parece?•
that sounds like a good idea — eso parece buena idea4.CPDsound archive N — archivo m de sonido
sound barrier N — barrera f del sonido
sound bite N — cita f jugosa
sound card N — (Comput) tarjeta f de sonido
sound effect N — efecto m sonoro
sound engineer N — ingeniero(-a) m / f de sonido
sound file N — (Comput) fichero m de sonido
sound library N — fonoteca f
sound mixer N — (=engineer) ingeniero(-a) m / f de sonido
sound recording N — grabación f sonora
sound recordist N — (TV) registrador(a) m / f de sonido
sound shift N — cambio m de pronunciación
sound system N — (Ling) sistema m fonológico; (=hi-fi) cadena f de sonido
sound truck N — (US) furgón m publicitario
sound wave N — (Phys) onda f sonora
II
[saʊnd]VT1) (Naut) sondar2) (Med) [+ chest] auscultar; [+ cavity, passage] sondar
III [saʊnd]1. ADJ(compar sounder) (superl soundest)1) (=in good condition) sano; [constitution] robusto; [structure] sólido, firme- be as sound as a bellsafe 1., 1)2) (=well-founded) [argument] bien fundado, sólido; [ideas, opinions] válido, razonable; [investment] bueno, seguro; [training] sólido; [decision, choice] acertado3) (=dependable) [person] formal, digno de confianzahe's a very sound man — es un hombre formal or digno de confianza
he's a sound worker — es buen trabajador, trabaja con seriedad
4) (=thorough)5) (=deep, untroubled) [sleep] profundo2.ADV
IV
[saʊnd]N (Geog) estrecho m, brazo m de mar* * *
I [saʊnd]1) noun2) u ca) ( noise) sonido m; (unpleasant, disturbing) ruido mb) (of music, instrument) sonido mc) ( Ling) sonido m3) ua) ( Phys) sonido m; (before n)b) (Audio, Rad, TV) sonido mturn the sound up/down — sube/baja el volumen; (before n)
sound effects — efectos mpl sonoros
4) ( impression conveyed) (colloq) (no pl)by o from the sound of it, everything's going very well — parece que or por lo visto todo marcha muy bien
5) ca) ( channel) paso m, estrecho mb) ( inlet) brazo m
II
1.
1)a) ( give impression) sonar*your voice sounds o you sound different on the phone — tu voz suena distinta por teléfono
you sound as if o as though you could do with a rest — me da la impresión de que no te vendría mal un descanso
it sounds as if o as though they're here now — (por el ruido) parece que ya están aquí
b) ( seem) parecer*we'll leave at ten; how does that sound to you? — saldremos a las diez ¿qué te parece?
it sounds as if o as though you had a great time — parece que lo pasaste fenomenal
sounds like fun! — (colloq) qué divertido!
2) (make noise, resound) \<\<bell/alarm\>\> sonar*
2.
vt1)a) \<\<trumpet/horn\>\> tocar*, hacer* sonarthe chairman sounded a note of warning in his speech — en su discurso, el presidente llamó a la cautela
b) ( articulate) \<\<letter/consonant\>\> pronunciar2) sound out•Phrasal Verbs:
III
adjective -er, -est1)a) ( healthy) sanoI, Peter Smith, being of sound mind... — (frml) yo, Peter Smith, (estando) en pleno uso de mis facultades... (frml)
b) ( in good condition) <basis/foundation> sólido, firme; < timber> en buenas condiciones2)a) ( valid) <reasoning/knowledge> sólido; <advice/decision> sensatob) ( reliable) <colleague/staff> responsable, formal3)b) (hard, thorough)
IV
adverb -er, -est -
19 logic
nounLogik, die* * *['lo‹ik]- academic.ru/43594/logical">logical- logically* * *log·ic[ˈlɒʤɪk, AM ˈlɑ:ʤ-]n no plhe cannot refuse to acknowledge the force of \logic er kann sich der zwingenden Logik nicht widersetzenthat's just \logic das ist einfach logischdeductive \logic logische Schlussfolgerungflawed \logic unlogischer Gedankenganginternal \logic innere Logikto defy \logic gegen jede Logik verstoßento follow sb's \logic jds Logik [o Gedankengang] folgenformal \logic die formale Logikthere's no \logic in the decision die Entscheidung ist wider alle Vernunft* * *['lɒdZɪk]nLogik f* * *2. Logik f:female logic weibliche Logik;chop logic Haarspalterei treibenb) Notwendigkeit f, Folgerichtigkeit f (einer Entwicklung etc)3. Überzeugungskraft f (von Fakten etc)* * *nounLogik, die* * *adj.logisch adj. n.Folgerichtigkeit f.Logik nur sing. f. -
20 Bibliography
■ Aitchison, J. (1987). Noam Chomsky: Consensus and controversy. New York: Falmer Press.■ Anderson, J. R. (1980). Cognitive psychology and its implications. San Francisco: W. H. Freeman.■ Anderson, J. R. (1983). The architecture of cognition. Cambridge, MA: Harvard University Press.■ Anderson, J. R. (1995). Cognitive psychology and its implications (4th ed.). New York: W. H. Freeman.■ Archilochus (1971). In M. L. West (Ed.), Iambi et elegi graeci (Vol. 1). Oxford: Oxford University Press.■ Armstrong, D. M. (1990). The causal theory of the mind. In W. G. Lycan (Ed.), Mind and cognition: A reader (pp. 37-47). Cambridge, MA: Basil Blackwell. (Originally published in 1981 in The nature of mind and other essays, Ithaca, NY: University Press).■ Atkins, P. W. (1992). Creation revisited. Oxford: W. H. Freeman & Company.■ Austin, J. L. (1962). How to do things with words. Cambridge, MA: Harvard University Press.■ Bacon, F. (1878). Of the proficience and advancement of learning divine and human. In The works of Francis Bacon (Vol. 1). Cambridge, MA: Hurd & Houghton.■ Bacon, R. (1928). Opus majus (Vol. 2). R. B. Burke (Trans.). Philadelphia, PA: University of Pennsylvania Press.■ Bar-Hillel, Y. (1960). The present status of automatic translation of languages. In F. L. Alt (Ed.), Advances in computers (Vol. 1). New York: Academic Press.■ Barr, A., & E. A. Feigenbaum (Eds.) (1981). The handbook of artificial intelligence (Vol. 1). Reading, MA: Addison-Wesley.■ Barr, A., & E. A. Feigenbaum (Eds.) (1982). The handbook of artificial intelligence (Vol. 2). Los Altos, CA: William Kaufman.■ Barron, F. X. (1963). The needs for order and for disorder as motives in creative activity. In C. W. Taylor & F. X. Barron (Eds.), Scientific creativity: Its rec ognition and development (pp. 153-160). New York: Wiley.■ Bartlett, F. C. (1932). Remembering: A study in experimental and social psychology. Cambridge: Cambridge University Press.■ Bartley, S. H. (1969). Principles of perception. London: Harper & Row.■ Barzun, J. (1959). The house of intellect. New York: Harper & Row.■ Beach, F. A., D. O. Hebb, C. T. Morgan & H. W. Nissen (Eds.) (1960). The neu ropsychology of Lashley. New York: McGraw-Hill.■ Berkeley, G. (1996). Principles of human knowledge: Three Dialogues. Oxford: Oxford University Press. (Originally published in 1710.)■ Berlin, I. (1953). The hedgehog and the fox: An essay on Tolstoy's view of history. NY: Simon & Schuster.■ Bierwisch, J. (1970). Semantics. In J. Lyons (Ed.), New horizons in linguistics. Baltimore: Penguin Books.■ Black, H. C. (1951). Black's law dictionary. St. Paul, MN: West Publishing.■ Bloom, A. (1981). The linguistic shaping of thought: A study in the impact of language on thinking in China and the West. Hillsdale, NJ: Erlbaum.■ Bobrow, D. G., & D. A. Norman (1975). Some principles of memory schemata. In D. G. Bobrow & A. Collins (Eds.), Representation and understanding: Stud ies in Cognitive Science (pp. 131-149). New York: Academic Press.■ Boden, M. A. (1977). Artificial intelligence and natural man. New York: Basic Books.■ Boden, M. A. (1981). Minds and mechanisms. Ithaca, NY: Cornell University Press.■ Boden, M. A. (1990a). The creative mind: Myths and mechanisms. London: Cardinal.■ Boden, M. A. (1990b). The philosophy of artificial intelligence. Oxford: Oxford University Press.■ Boden, M. A. (1994). Precis of The creative mind: Myths and mechanisms. Behavioral and brain sciences 17, 519-570.■ Boden, M. (1996). Creativity. In M. Boden (Ed.), Artificial Intelligence (2nd ed.). San Diego: Academic Press.■ Bolter, J. D. (1984). Turing's man: Western culture in the computer age. Chapel Hill, NC: University of North Carolina Press.■ Bolton, N. (1972). The psychology of thinking. London: Methuen.■ Bourne, L. E. (1973). Some forms of cognition: A critical analysis of several papers. In R. Solso (Ed.), Contemporary issues in cognitive psychology (pp. 313324). Loyola Symposium on Cognitive Psychology (Chicago 1972). Washington, DC: Winston.■ Bransford, J. D., N. S. McCarrell, J. J. Franks & K. E. Nitsch (1977). Toward unexplaining memory. In R. Shaw & J. D. Bransford (Eds.), Perceiving, acting, and knowing (pp. 431-466). Hillsdale, NJ: Lawrence Erlbaum Associates.■ Breger, L. (1981). Freud's unfinished journey. London: Routledge & Kegan Paul.■ Brehmer, B. (1986). In one word: Not from experience. In H. R. Arkes & K. Hammond (Eds.), Judgment and decision making: An interdisciplinary reader (pp. 705-719). Cambridge: Cambridge University Press.■ Bresnan, J. (1978). A realistic transformational grammar. In M. Halle, J. Bresnan & G. A. Miller (Eds.), Linguistic theory and psychological reality (pp. 1-59). Cambridge, MA: MIT Press.■ Brislin, R. W., W. J. Lonner & R. M. Thorndike (Eds.) (1973). Cross- cultural research methods. New York: Wiley.■ Bronowski, J. (1977). A sense of the future: Essays in natural philosophy. P. E. Ariotti with R. Bronowski (Eds.). Cambridge, MA: MIT Press.■ Bronowski, J. (1978). The origins of knowledge and imagination. New Haven, CT: Yale University Press.■ Brown, R. O. (1973). A first language: The early stages. Cambridge, MA: Harvard University Press.■ Brown, T. (1970). Lectures on the philosophy of the human mind. In R. Brown (Ed.), Between Hume and Mill: An anthology of British philosophy- 1749- 1843 (pp. 330-387). New York: Random House/Modern Library.■ Bruner, J. S., J. Goodnow & G. Austin (1956). A study of thinking. New York: Wiley.■ Calvin, W. H. (1990). The cerebral symphony: Seashore reflections on the structure of consciousness. New York: Bantam.■ Campbell, J. (1982). Grammatical man: Information, entropy, language, and life. New York: Simon & Schuster.■ Campbell, J. (1989). The improbable machine. New York: Simon & Schuster.■ Carlyle, T. (1966). On heroes, hero- worship and the heroic in history. Lincoln: University of Nebraska Press. (Originally published in 1841.)■ Carnap, R. (1959). The elimination of metaphysics through logical analysis of language [Ueberwindung der Metaphysik durch logische Analyse der Sprache]. In A. J. Ayer (Ed.), Logical positivism (pp. 60-81) A. Pap (Trans). New York: Free Press. (Originally published in 1932.)■ Cassirer, E. (1946). Language and myth. New York: Harper and Brothers. Reprinted. New York: Dover Publications, 1953.■ Cattell, R. B., & H. J. Butcher (1970). Creativity and personality. In P. E. Vernon (Ed.), Creativity. Harmondsworth, England: Penguin Books.■ Caudill, M., & C. Butler (1990). Naturally intelligent systems. Cambridge, MA: MIT Press/Bradford Books.■ Chandrasekaran, B. (1990). What kind of information processing is intelligence? A perspective on AI paradigms and a proposal. In D. Partridge & R. Wilks (Eds.), The foundations of artificial intelligence: A sourcebook (pp. 14-46). Cambridge: Cambridge University Press.■ Charniak, E., & McDermott, D. (1985). Introduction to artificial intelligence. Reading, MA: Addison-Wesley.■ Chase, W. G., & H. A. Simon (1988). The mind's eye in chess. In A. Collins & E. E. Smith (Eds.), Readings in cognitive science: A perspective from psychology and artificial intelligence (pp. 461-493). San Mateo, CA: Kaufmann.■ Cheney, D. L., & R. M. Seyfarth (1990). How monkeys see the world: Inside the mind of another species. Chicago: University of Chicago Press.■ Chi, M.T.H., R. Glaser & E. Rees (1982). Expertise in problem solving. In R. J. Sternberg (Ed.), Advances in the psychology of human intelligence (pp. 7-73). Hillsdale, NJ: Lawrence Erlbaum Associates.■ Chomsky, N. (1957). Syntactic structures. The Hague: Mouton. Janua Linguarum.■ Chomsky, N. (1964). A transformational approach to syntax. In J. A. Fodor & J. J. Katz (Eds.), The structure of language: Readings in the philosophy of lan guage (pp. 211-245). Englewood Cliffs, NJ: Prentice-Hall.■ Chomsky, N. (1965). Aspects of the theory of syntax. Cambridge, MA: MIT Press.■ Chomsky, N. (1972). Language and mind (enlarged ed.). New York: Harcourt Brace Jovanovich.■ Chomsky, N. (1979). Language and responsibility. New York: Pantheon.■ Chomsky, N. (1986). Knowledge of language: Its nature, origin and use. New York: Praeger Special Studies.■ Churchland, P. (1979). Scientific realism and the plasticity of mind. New York: Cambridge University Press.■ Churchland, P. M. (1989). A neurocomputational perspective: The nature of mind and the structure of science. Cambridge, MA: MIT Press.■ Churchland, P. S. (1986). Neurophilosophy. Cambridge, MA: MIT Press/Bradford Books.■ Clark, A. (1996). Philosophical Foundations. In M. A. Boden (Ed.), Artificial in telligence (2nd ed.). San Diego: Academic Press.■ Clark, H. H., & T. B. Carlson (1981). Context for comprehension. In J. Long & A. Baddeley (Eds.), Attention and performance (Vol. 9, pp. 313-330). Hillsdale, NJ: Lawrence Erlbaum Associates.■ Clarke, A. C. (1984). Profiles of the future: An inquiry into the limits of the possible. New York: Holt, Rinehart & Winston.■ Claxton, G. (1980). Cognitive psychology: A suitable case for what sort of treatment? In G. Claxton (Ed.), Cognitive psychology: New directions (pp. 1-25). London: Routledge & Kegan Paul.■ Code, M. (1985). Order and organism. Albany, NY: State University of New York Press.■ Collingwood, R. G. (1972). The idea of history. New York: Oxford University Press.■ Coopersmith, S. (1967). The antecedents of self- esteem. San Francisco: W. H. Freeman.■ Copland, A. (1952). Music and imagination. London: Oxford University Press.■ Coren, S. (1994). The intelligence of dogs. New York: Bantam Books.■ Cottingham, J. (Ed.) (1996). Western philosophy: An anthology. Oxford: Blackwell Publishers.■ Cox, C. (1926). The early mental traits of three hundred geniuses. Stanford, CA: Stanford University Press.■ Craik, K.J.W. (1943). The nature of explanation. Cambridge: Cambridge University Press.■ Cronbach, L. J. (1990). Essentials of psychological testing (5th ed.). New York: HarperCollins.■ Cronbach, L. J., & R. E. Snow (1977). Aptitudes and instructional methods. New York: Irvington. Paperback edition, 1981.■ Csikszentmihalyi, M. (1993). The evolving self. New York: Harper Perennial.■ Culler, J. (1976). Ferdinand de Saussure. New York: Penguin Books.■ Curtius, E. R. (1973). European literature and the Latin Middle Ages. W. R. Trask (Trans.). Princeton, NJ: Princeton University Press.■ D'Alembert, J.L.R. (1963). Preliminary discourse to the encyclopedia of Diderot. R. N. Schwab (Trans.). Indianapolis: Bobbs-Merrill.■ Dampier, W. C. (1966). A history of modern science. Cambridge: Cambridge University Press.■ Darwin, C. (1911). The life and letters of Charles Darwin (Vol. 1). Francis Darwin (Ed.). New York: Appleton.■ Davidson, D. (1970) Mental events. In L. Foster & J. W. Swanson (Eds.), Experience and theory (pp. 79-101). Amherst: University of Massachussetts Press.■ Davies, P. (1995). About time: Einstein's unfinished revolution. New York: Simon & Schuster/Touchstone.■ Davis, R., & J. J. King (1977). An overview of production systems. In E. Elcock & D. Michie (Eds.), Machine intelligence 8. Chichester, England: Ellis Horwood.■ Davis, R., & D. B. Lenat (1982). Knowledge- based systems in artificial intelligence. New York: McGraw-Hill.■ Dawkins, R. (1982). The extended phenotype: The gene as the unit of selection. Oxford: W. H. Freeman.■ deKleer, J., & J. S. Brown (1983). Assumptions and ambiguities in mechanistic mental models (1983). In D. Gentner & A. L. Stevens (Eds.), Mental modes (pp. 155-190). Hillsdale, NJ: Lawrence Erlbaum Associates.■ Dennett, D. C. (1978a). Brainstorms: Philosophical essays on mind and psychology. Montgomery, VT: Bradford Books.■ Dennett, D. C. (1978b). Toward a cognitive theory of consciousness. In D. C. Dennett, Brainstorms: Philosophical Essays on Mind and Psychology. Montgomery, VT: Bradford Books.■ Dennett, D. C. (1995). Darwin's dangerous idea: Evolution and the meanings of life. New York: Simon & Schuster/Touchstone.■ Descartes, R. (1897-1910). Traite de l'homme. In Oeuvres de Descartes (Vol. 11, pp. 119-215). Paris: Charles Adam & Paul Tannery. (Originally published in 1634.)■ Descartes, R. (1950). Discourse on method. L. J. Lafleur (Trans.). New York: Liberal Arts Press. (Originally published in 1637.)■ Descartes, R. (1951). Meditation on first philosophy. L. J. Lafleur (Trans.). New York: Liberal Arts Press. (Originally published in 1641.)■ Descartes, R. (1955). The philosophical works of Descartes. E. S. Haldane and G.R.T. Ross (Trans.). New York: Dover. (Originally published in 1911 by Cambridge University Press.)■ Descartes, R. (1967). Discourse on method (Pt. V). In E. S. Haldane and G.R.T. Ross (Eds.), The philosophical works of Descartes (Vol. 1, pp. 106-118). Cambridge: Cambridge University Press. (Originally published in 1637.)■ Descartes, R. (1970a). Discourse on method. In E. S. Haldane & G.R.T. Ross (Eds.), The philosophical works of Descartes (Vol. 1, pp. 181-200). Cambridge: Cambridge University Press. (Originally published in 1637.)■ Descartes, R. (1970b). Principles of philosophy. In E. S. Haldane & G.R.T. Ross (Eds.), The philosophical works of Descartes (Vol. 1, pp. 178-291). Cambridge: Cambridge University Press. (Originally published in 1644.)■ Descartes, R. (1984). Meditations on first philosophy. In J. Cottingham, R. Stoothoff & D. Murduch (Trans.), The philosophical works of Descartes (Vol. 2). Cambridge: Cambridge University Press. (Originally published in 1641.)■ Descartes, R. (1986). Meditations on first philosophy. J. Cottingham (Trans.). Cambridge: Cambridge University Press. (Originally published in 1641 as Med itationes de prima philosophia.)■ deWulf, M. (1956). An introduction to scholastic philosophy. Mineola, NY: Dover Books.■ Dixon, N. F. (1981). Preconscious processing. London: Wiley.■ Doyle, A. C. (1986). The Boscombe Valley mystery. In Sherlock Holmes: The com plete novels and stories (Vol. 1). New York: Bantam.■ Dreyfus, H., & S. Dreyfus (1986). Mind over machine. New York: Free Press.■ Dreyfus, H. L. (1972). What computers can't do: The limits of artificial intelligence (revised ed.). New York: Harper & Row.■ Dreyfus, H. L., & S. E. Dreyfus (1986). Mind over machine: The power of human intuition and expertise in the era of the computer. New York: Free Press.■ Edelman, G. M. (1992). Bright air, brilliant fire: On the matter of the mind. New York: Basic Books.■ Ehrenzweig, A. (1967). The hidden order of art. London: Weidenfeld & Nicolson.■ Einstein, A., & L. Infeld (1938). The evolution of physics. New York: Simon & Schuster.■ Eisenstein, S. (1947). Film sense. New York: Harcourt, Brace & World.■ Everdell, W. R. (1997). The first moderns. Chicago: University of Chicago Press.■ Eysenck, M. W. (1977). Human memory: Theory, research and individual difference. Oxford: Pergamon.■ Eysenck, M. W. (1982). Attention and arousal: Cognition and performance. Berlin: Springer.■ Eysenck, M. W. (1984). A handbook of cognitive psychology. Hillsdale, NJ: Lawrence Erlbaum Associates.■ Fancher, R. E. (1979). Pioneers of psychology. New York: W. W. Norton.■ Farrell, B. A. (1981). The standing of psychoanalysis. New York: Oxford University Press.■ Feldman, D. H. (1980). Beyond universals in cognitive development. Norwood, NJ: Ablex.■ Fetzer, J. H. (1996). Philosophy and cognitive science (2nd ed.). New York: Paragon House.■ Finke, R. A. (1990). Creative imagery: Discoveries and inventions in visualization. Hillsdale, NJ: Lawrence Erlbaum Associates.■ Flanagan, O. (1991). The science of the mind. Cambridge MA: MIT Press/Bradford Books.■ Fodor, J. (1983). The modularity of mind. Cambridge, MA: MIT Press/Bradford Books.■ Frege, G. (1972). Conceptual notation. T. W. Bynum (Trans.). Oxford: Clarendon Press. (Originally published in 1879.)■ Frege, G. (1979). Logic. In H. Hermes, F. Kambartel & F. Kaulbach (Eds.), Gottlob Frege: Posthumous writings. Chicago: University of Chicago Press. (Originally published in 1879-1891.)■ Freud, S. (1959). Creative writers and day-dreaming. In J. Strachey (Ed.), The standard edition of the complete psychological works of Sigmund Freud (Vol. 9, pp. 143-153). London: Hogarth Press.■ Freud, S. (1966). Project for a scientific psychology. In J. Strachey (Ed.), The stan dard edition of the complete psychological works of Sigmund Freud (Vol. 1, pp. 295-398). London: Hogarth Press. (Originally published in 1950 as Aus den AnfaЁngen der Psychoanalyse, in London by Imago Publishing.)■ Freud, S. (1976). Lecture 18-Fixation to traumas-the unconscious. In J. Strachey (Ed.), The standard edition of the complete psychological works of Sigmund Freud (Vol. 16, p. 285). London: Hogarth Press.■ Galileo, G. (1990). Il saggiatore [The assayer]. In S. Drake (Ed.), Discoveries and opinions of Galileo. New York: Anchor Books. (Originally published in 1623.)■ Gassendi, P. (1970). Letter to Descartes. In "Objections and replies." In E. S. Haldane & G.R.T. Ross (Eds.), The philosophical works of Descartes (Vol. 2, pp. 179-240). Cambridge: Cambridge University Press. (Originally published in 1641.)■ Gazzaniga, M. S. (1988). Mind matters: How mind and brain interact to create our conscious lives. Boston: Houghton Mifflin in association with MIT Press/Bradford Books.■ Genesereth, M. R., & N. J. Nilsson (1987). Logical foundations of artificial intelligence. Palo Alto, CA: Morgan Kaufmann.■ Ghiselin, B. (1952). The creative process. New York: Mentor.■ Ghiselin, B. (1985). The creative process. Berkeley, CA: University of California Press. (Originally published in 1952.)■ Gilhooly, K. J. (1996). Thinking: Directed, undirected and creative (3rd ed.). London: Academic Press.■ Glass, A. L., K. J. Holyoak & J. L. Santa (1979). Cognition. Reading, MA: AddisonWesley.■ Goody, J. (1977). The domestication of the savage mind. Cambridge: Cambridge University Press.■ Gruber, H. E. (1980). Darwin on man: A psychological study of scientific creativity (2nd ed.). Chicago: University of Chicago Press.■ Gruber, H. E., & S. Davis (1988). Inching our way up Mount Olympus: The evolving systems approach to creative thinking. In R. J. Sternberg (Ed.), The nature of creativity: Contemporary psychological perspectives. Cambridge: Cambridge University Press.■ Guthrie, E. R. (1972). The psychology of learning. New York: Harper. (Originally published in 1935.)■ Habermas, J. (1972). Knowledge and human interests. Boston: Beacon Press.■ Hadamard, J. (1945). The psychology of invention in the mathematical field. Princeton, NJ: Princeton University Press.■ Hand, D. J. (1985). Artificial intelligence and psychiatry. Cambridge: Cambridge University Press.■ Harris, M. (1981). The language myth. London: Duckworth.■ Haugeland, J. (Ed.) (1981). Mind design: Philosophy, psychology, artificial intelligence. Cambridge, MA: MIT Press/Bradford Books.■ Haugeland, J. (1981a). The nature and plausibility of cognitivism. In J. Haugeland (Ed.), Mind design: Philosophy, psychology, artificial intelligence (pp. 243-281). Cambridge, MA: MIT Press.■ Haugeland, J. (1981b). Semantic engines: An introduction to mind design. In J. Haugeland (Ed.), Mind design: Philosophy, psychology, artificial intelligence (pp. 1-34). Cambridge, MA: MIT Press/Bradford Books.■ Haugeland, J. (1985). Artificial intelligence: The very idea. Cambridge, MA: MIT Press.■ Hawkes, T. (1977). Structuralism and semiotics. Berkeley: University of California Press.■ Hebb, D. O. (1949). The organisation of behaviour. New York: Wiley.■ Hebb, D. O. (1958). A textbook of psychology. Philadelphia: Saunders.■ Hegel, G.W.F. (1910). The phenomenology of mind. J. B. Baille (Trans.). London: Sonnenschein. (Originally published as Phaenomenologie des Geistes, 1807.)■ Heisenberg, W. (1958). Physics and philosophy. New York: Harper & Row.■ Hempel, C. G. (1966). Philosophy of natural science. Englewood Cliffs, NJ: PrenticeHall.■ Herman, A. (1997). The idea of decline in Western history. New York: Free Press.■ Herrnstein, R. J., & E. G. Boring (Eds.) (1965). A source book in the history of psy chology. Cambridge, MA: Harvard University Press.■ Herzmann, E. (1964). Mozart's creative process. In P. H. Lang (Ed.), The creative world of Mozart (pp. 17-30). London: Oldbourne Press.■ Hilgard, E. R. (1957). Introduction to psychology. London: Methuen.■ Hobbes, T. (1651). Leviathan. London: Crooke.■ Holliday, S. G., & M. J. Chandler (1986). Wisdom: Explorations in adult competence. Basel, Switzerland: Karger.■ Horn, J. L. (1986). In R. J. Sternberg (Ed.), Advances in the psychology of human intelligence (Vol. 3). Hillsdale, NJ: Erlbaum.■ Hull, C. (1943). Principles of behavior. New York: Appleton-Century-Crofts.■ Hume, D. (1955). An inquiry concerning human understanding. New York: Liberal Arts Press. (Originally published in 1748.)■ Hume, D. (1975). An enquiry concerning human understanding. In L. A. SelbyBigge (Ed.), Hume's enquiries (3rd. ed., revised P. H. Nidditch). Oxford: Clarendon. (Spelling and punctuation revised.) (Originally published in 1748.)■ Hume, D. (1978). A treatise of human nature. L. A. Selby-Bigge (Ed.), Hume's enquiries (3rd. ed., revised P. H. Nidditch). Oxford: Clarendon. (With some modifications of spelling and punctuation.) (Originally published in 1690.)■ Hunt, E. (1973). The memory we must have. In R. C. Schank & K. M. Colby (Eds.), Computer models of thought and language. (pp. 343-371) San Francisco: W. H. Freeman.■ Husserl, E. (1960). Cartesian meditations. The Hague: Martinus Nijhoff.■ Inhelder, B., & J. Piaget (1958). The growth of logical thinking from childhood to adolescence. New York: Basic Books. (Originally published in 1955 as De la logique de l'enfant a` la logique de l'adolescent. [Paris: Presses Universitaire de France])■ James, W. (1890a). The principles of psychology (Vol. 1). New York: Dover Books.■ James, W. (1890b). The principles of psychology. New York: Henry Holt.■ Jevons, W. S. (1900). The principles of science (2nd ed.). London: Macmillan.■ Johnson, G. (1986). Machinery of the mind: Inside the new science of artificial intelli gence. New York: Random House.■ Johnson-Laird, P. N. (1983). Mental models: Toward a cognitive science of language, inference, and consciousness. Cambridge, MA: Harvard University Press.■ Johnson-Laird, P. N. (1988). The computer and the mind: An introduction to cognitive science. Cambridge, MA: Harvard University Press.■ Jones, E. (1961). The life and work of Sigmund Freud. L. Trilling & S. Marcus (Eds.). London: Hogarth.■ Jones, R. V. (1985). Complementarity as a way of life. In A. P. French & P. J. Kennedy (Eds.), Niels Bohr: A centenary volume. Cambridge, MA: Harvard University Press.■ Kant, I. (1933). Critique of Pure Reason (2nd ed.). N. K. Smith (Trans.). London: Macmillan. (Originally published in 1781 as Kritik der reinen Vernunft.)■ Kant, I. (1891). Solution of the general problems of the Prolegomena. In E. Belfort (Trans.), Kant's Prolegomena. London: Bell. (With minor modifications.) (Originally published in 1783.)■ Katona, G. (1940). Organizing and memorizing: Studies in the psychology of learning and teaching. New York: Columbia University Press.■ Kaufman, A. S. (1979). Intelligent testing with the WISC-R. New York: Wiley.■ Koestler, A. (1964). The act of creation. New York: Arkana (Penguin).■ Kohlberg, L. (1971). From is to ought. In T. Mischel (Ed.), Cognitive development and epistemology. (pp. 151-235) New York: Academic Press.■ KoЁhler, W. (1925). The mentality of apes. New York: Liveright.■ KoЁhler, W. (1927). The mentality of apes (2nd ed.). Ella Winter (Trans.). London: Routledge & Kegan Paul.■ KoЁhler, W. (1930). Gestalt psychology. London: G. Bell.■ KoЁhler, W. (1947). Gestalt psychology. New York: Liveright.■ KoЁhler, W. (1969). The task of Gestalt psychology. Princeton, NJ: Princeton University Press.■ Kuhn, T. (1970). The structure of scientific revolutions (2nd ed.). Chicago: University of Chicago Press.■ Langer, E. J. (1989). Mindfulness. Reading, MA: Addison-Wesley.■ Langer, S. (1962). Philosophical sketches. Baltimore: Johns Hopkins University Press.■ Langley, P., H. A. Simon, G. L. Bradshaw & J. M. Zytkow (1987). Scientific dis covery: Computational explorations of the creative process. Cambridge, MA: MIT Press.■ Lashley, K. S. (1951). The problem of serial order in behavior. In L. A. Jeffress (Ed.), Cerebral mechanisms in behavior, the Hixon Symposium (pp. 112-146) New York: Wiley.■ LeDoux, J. E., & W. Hirst (1986). Mind and brain: Dialogues in cognitive neuroscience. Cambridge: Cambridge University Press.■ Lehnert, W. (1978). The process of question answering. Hillsdale, NJ: Lawrence Erlbaum Associates.■ Leiber, J. (1991). Invitation to cognitive science. Oxford: Blackwell.■ Lenat, D. B., & G. Harris (1978). Designing a rule system that searches for scientific discoveries. In D. A. Waterman & F. Hayes-Roth (Eds.), Pattern directed inference systems (pp. 25-52) New York: Academic Press.■ Levenson, T. (1995). Measure for measure: A musical history of science. New York: Touchstone. (Originally published in 1994.)■ Leґvi-Strauss, C. (1963). Structural anthropology. C. Jacobson & B. Grundfest Schoepf (Trans.). New York: Basic Books. (Originally published in 1958.)■ Levine, M. W., & J. M. Schefner (1981). Fundamentals of sensation and perception. London: Addison-Wesley.■ Lewis, C. I. (1946). An analysis of knowledge and valuation. LaSalle, IL: Open Court.■ Lighthill, J. (1972). A report on artificial intelligence. Unpublished manuscript, Science Research Council.■ Lipman, M., A. M. Sharp & F. S. Oscanyan (1980). Philosophy in the classroom. Philadelphia: Temple University Press.■ Lippmann, W. (1965). Public opinion. New York: Free Press. (Originally published in 1922.)■ Locke, J. (1956). An essay concerning human understanding. Chicago: Henry Regnery Co. (Originally published in 1690.)■ Locke, J. (1975). An essay concerning human understanding. P. H. Nidditch (Ed.). Oxford: Clarendon. (Originally published in 1690.) (With spelling and punctuation modernized and some minor modifications of phrasing.)■ Lopate, P. (1994). The art of the personal essay. New York: Doubleday/Anchor Books.■ Lorimer, F. (1929). The growth of reason. London: Kegan Paul. Machlup, F., & U. Mansfield (Eds.) (1983). The study of information. New York: Wiley.■ Manguel, A. (1996). A history of reading. New York: Viking.■ Markey, J. F. (1928). The symbolic process. London: Kegan Paul.■ Martin, R. M. (1969). On Ziff's "Natural and formal languages." In S. Hook (Ed.), Language and philosophy: A symposium (pp. 249-263). New York: New York University Press.■ Mazlish, B. (1993). The fourth discontinuity: the co- evolution of humans and machines. New Haven, CT: Yale University Press.■ McCarthy, J., & P. J. Hayes (1969). Some philosophical problems from the standpoint of artificial intelligence. In B. Meltzer & D. Michie (Eds.), Machine intelligence 4. Edinburgh: Edinburgh University Press.■ McClelland, J. L., D. E. Rumelhart & G. E. Hinton (1986). The appeal of parallel distributed processing. In D. E. Rumelhart, J. L. McClelland & the PDP Research Group (Eds.), Parallel distributed processing: Explorations in the mi crostructure of cognition (Vol. 1, pp. 3-40). Cambridge, MA: MIT Press/ Bradford Books.■ McCorduck, P. (1979). Machines who think. San Francisco: W. H. Freeman.■ McLaughlin, T. (1970). Music and communication. London: Faber & Faber.■ Mednick, S. A. (1962). The associative basis of the creative process. Psychological Review 69, 431-436.■ Meehl, P. E., & C. J. Golden (1982). Taxometric methods. In Kendall, P. C., & Butcher, J. N. (Eds.), Handbook of research methods in clinical psychology (pp. 127-182). New York: Wiley.■ Mehler, J., E.C.T. Walker & M. Garrett (Eds.) (1982). Perspectives on mental rep resentation: Experimental and theoretical studies of cognitive processes and ca pacities. Hillsdale, NJ: Lawrence Erlbaum Associates.■ Mill, J. S. (1900). A system of logic, ratiocinative and inductive: Being a connected view of the principles of evidence and the methods of scientific investigation. London: Longmans, Green.■ Miller, G. A. (1979, June). A very personal history. Talk to the Cognitive Science Workshop, Cambridge, MA.■ Miller, J. (1983). States of mind. New York: Pantheon Books.■ Minsky, M. (1975). A framework for representing knowledge. In P. H. Winston (Ed.), The psychology of computer vision (pp. 211-277). New York: McGrawHill.■ Minsky, M., & S. Papert (1973). Artificial intelligence. Condon Lectures, Oregon State System of Higher Education, Eugene, Oregon.■ Minsky, M. L. (1986). The society of mind. New York: Simon & Schuster.■ Mischel, T. (1976). Psychological explanations and their vicissitudes. In J. K. Cole & W. J. Arnold (Eds.), Nebraska Symposium on motivation (Vol. 23). Lincoln, NB: University of Nebraska Press.■ Morford, M.P.O., & R. J. Lenardon (1995). Classical mythology (5th ed.). New York: Longman.■ Murdoch, I. (1954). Under the net. New York: Penguin.■ Nagel, E. (1959). Methodological issues in psychoanalytic theory. In S. Hook (Ed.), Psychoanalysis, scientific method, and philosophy: A symposium. New York: New York University Press.■ Nagel, T. (1979). Mortal questions. London: Cambridge University Press.■ Nagel, T. (1986). The view from nowhere. Oxford: Oxford University Press.■ Neisser, U. (1967). Cognitive psychology. New York: Appleton-Century-Crofts.■ Neisser, U. (1972). Changing conceptions of imagery. In P. W. Sheehan (Ed.), The function and nature of imagery (pp. 233-251). London: Academic Press.■ Neisser, U. (1976). Cognition and reality. San Francisco: W. H. Freeman.■ Neisser, U. (1978). Memory: What are the important questions? In M. M. Gruneberg, P. E. Morris & R. N. Sykes (Eds.), Practical aspects of memory (pp. 3-24). London: Academic Press.■ Neisser, U. (1979). The concept of intelligence. In R. J. Sternberg & D. K. Detterman (Eds.), Human intelligence: Perspectives on its theory and measurement (pp. 179-190). Norwood, NJ: Ablex.■ Nersessian, N. (1992). How do scientists think? Capturing the dynamics of conceptual change in science. In R. N. Giere (Ed.), Cognitive models of science (pp. 3-44). Minneapolis: University of Minnesota Press.■ Newell, A. (1973a). Artificial intelligence and the concept of mind. In R. C. Schank & K. M. Colby (Eds.), Computer models of thought and language (pp. 1-60). San Francisco: W. H. Freeman.■ Newell, A. (1973b). You can't play 20 questions with nature and win. In W. G. Chase (Ed.), Visual information processing (pp. 283-310). New York: Academic Press.■ Newell, A., & H. A. Simon (1963). GPS: A program that simulates human thought. In E. A. Feigenbaum & J. Feldman (Eds.), Computers and thought (pp. 279-293). New York & McGraw-Hill.■ Newell, A., & H. A. Simon (1972). Human problem solving. Englewood Cliffs, NJ: Prentice-Hall.■ Nietzsche, F. (1966). Beyond good and evil. W. Kaufmann (Trans.). New York: Vintage. (Originally published in 1885.)■ Nilsson, N. J. (1971). Problem- solving methods in artificial intelligence. New York: McGraw-Hill.■ Nussbaum, M. C. (1978). Aristotle's Princeton University Press. De Motu Anamalium. Princeton, NJ:■ Oersted, H. C. (1920). Thermo-electricity. In Kirstine Meyer (Ed.), H. C. Oersted, Natuurvidenskabelige Skrifter (Vol. 2). Copenhagen: n.p. (Originally published in 1830 in The Edinburgh encyclopaedia.)■ Ong, W. J. (1982). Orality and literacy: The technologizing of the word. London: Methuen.■ Onians, R. B. (1954). The origins of European thought. Cambridge, MA: Cambridge University Press.■ Osgood, C. E. (1960). Method and theory in experimental psychology. New York: Oxford University Press. (Originally published in 1953.)■ Osgood, C. E. (1966). Language universals and psycholinguistics. In J. H. Greenberg (Ed.), Universals of language (2nd ed., pp. 299-322). Cambridge, MA: MIT Press.■ Palmer, R. E. (1969). Hermeneutics. Evanston, IL: Northwestern University Press.■ Peirce, C. S. (1934). Some consequences of four incapacities-Man, a sign. In C. Hartsborne & P. Weiss (Eds.), Collected papers of Charles Saunders Peirce (Vol. 5, pp. 185-189). Cambridge, MA: Harvard University Press.■ Penfield, W. (1959). In W. Penfield & L. Roberts, Speech and brain mechanisms. Princeton, NJ: Princeton University Press.■ Penrose, R. (1994). Shadows of the mind: A search for the missing science of conscious ness. Oxford: Oxford University Press.■ Perkins, D. N. (1981). The mind's best work. Cambridge, MA: Harvard University Press.■ Peterfreund, E. (1986). The heuristic approach to psychoanalytic therapy. In■ J. Reppen (Ed.), Analysts at work, (pp. 127-144). Hillsdale, NJ: Analytic Press.■ Piaget, J. (1952). The origin of intelligence in children. New York: International Universities Press. (Originally published in 1936.)■ Piaget, J. (1954). Le langage et les opeґrations intellectuelles. Proble` mes de psycho linguistique. Symposium de l'Association de Psychologie Scientifique de Langue Francёaise. Paris: Presses Universitaires de France.■ Piaget, J. (1977). Problems of equilibration. In H. E. Gruber & J. J. Voneche (Eds.), The essential Piaget (pp. 838-841). London: Routlege & Kegan Paul. (Originally published in 1975 as L'eґquilibration des structures cognitives [Paris: Presses Universitaires de France].)■ Piaget, J., & B. Inhelder. (1973). Memory and intelligence. New York: Basic Books.■ Pinker, S. (1994). The language instinct. New York: Morrow.■ Pinker, S. (1996). Facts about human language relevant to its evolution. In J.-P. Changeux & J. Chavaillon (Eds.), Origins of the human brain. A symposium of the Fyssen foundation (pp. 262-283). Oxford: Clarendon Press. Planck, M. (1949). Scientific autobiography and other papers. F. Gaynor (Trans.). New York: Philosophical Library.■ Planck, M. (1990). Wissenschaftliche Selbstbiographie. W. Berg (Ed.). Halle, Germany: Deutsche Akademie der Naturforscher Leopoldina.■ Plato (1892). Meno. In The Dialogues of Plato (B. Jowett, Trans.; Vol. 2). New York: Clarendon. (Originally published circa 380 B.C.)■ Poincareґ, H. (1913). Mathematical creation. In The foundations of science. G. B. Halsted (Trans.). New York: Science Press.■ Poincareґ, H. (1921). The foundations of science: Science and hypothesis, the value of science, science and method. G. B. Halstead (Trans.). New York: Science Press.■ Poincareґ, H. (1929). The foundations of science: Science and hypothesis, the value of science, science and method. New York: Science Press.■ Poincareґ, H. (1952). Science and method. F. Maitland (Trans.) New York: Dover.■ Polya, G. (1945). How to solve it. Princeton, NJ: Princeton University Press.■ Polanyi, M. (1958). Personal knowledge. London: Routledge & Kegan Paul.■ Popper, K. (1968). Conjectures and refutations: The growth of scientific knowledge. New York: Harper & Row/Basic Books.■ Popper, K., & J. Eccles (1977). The self and its brain. New York: Springer-Verlag.■ Popper, K. R. (1959). The logic of scientific discovery. London: Hutchinson.■ Putnam, H. (1975). Mind, language and reality: Philosophical papers (Vol. 2). Cambridge: Cambridge University Press.■ Putnam, H. (1987). The faces of realism. LaSalle, IL: Open Court.■ Pylyshyn, Z. W. (1981). The imagery debate: Analog media versus tacit knowledge. In N. Block (Ed.), Imagery (pp. 151-206). Cambridge, MA: MIT Press.■ Pylyshyn, Z. W. (1984). Computation and cognition: Towards a foundation for cog nitive science. Cambridge, MA: MIT Press/Bradford Books.■ Quillian, M. R. (1968). Semantic memory. In M. Minsky (Ed.), Semantic information processing (pp. 216-260). Cambridge, MA: MIT Press.■ Quine, W.V.O. (1960). Word and object. Cambridge, MA: Harvard University Press.■ Rabbitt, P.M.A., & S. Dornic (Eds.). Attention and performance (Vol. 5). London: Academic Press.■ Rawlins, G.J.E. (1997). Slaves of the Machine: The quickening of computer technology. Cambridge, MA: MIT Press/Bradford Books.■ Reid, T. (1970). An inquiry into the human mind on the principles of common sense. In R. Brown (Ed.), Between Hume and Mill: An anthology of British philosophy- 1749- 1843 (pp. 151-178). New York: Random House/Modern Library.■ Reitman, W. (1970). What does it take to remember? In D. A. Norman (Ed.), Models of human memory (pp. 470-510). London: Academic Press.■ Ricoeur, P. (1974). Structure and hermeneutics. In D. I. Ihde (Ed.), The conflict of interpretations: Essays in hermeneutics (pp. 27-61). Evanston, IL: Northwestern University Press.■ Robinson, D. N. (1986). An intellectual history of psychology. Madison: University of Wisconsin Press.■ Rorty, R. (1979). Philosophy and the mirror of nature. Princeton, NJ: Princeton University Press.■ Rosch, E. (1977). Human categorization. In N. Warren (Ed.), Studies in cross cultural psychology (Vol. 1, pp. 1-49) London: Academic Press.■ Rosch, E. (1978). Principles of categorization. In E. Rosch & B. B. Lloyd (Eds.), Cognition and categorization (pp. 27-48). Hillsdale, NJ: Lawrence Erlbaum Associates.■ Rosch, E., & B. B. Lloyd (1978). Principles of categorization. In E. Rosch & B. B. Lloyd (Eds.), Cognition and categorization. Hillsdale, NJ: Lawrence Erlbaum Associates.■ Rose, S. (1970). The chemistry of life. Baltimore: Penguin Books.■ Rose, S. (1976). The conscious brain (updated ed.). New York: Random House.■ Rose, S. (1993). The making of memory: From molecules to mind. New York: Anchor Books. (Originally published in 1992)■ Roszak, T. (1994). The cult of information: A neo- Luddite treatise on high- tech, artificial intelligence, and the true art of thinking (2nd ed.). Berkeley: University of California Press.■ Royce, J. R., & W. W. Rozeboom (Eds.) (1972). The psychology of knowing. New York: Gordon & Breach.■ Rumelhart, D. E. (1977). Introduction to human information processing. New York: Wiley.■ Rumelhart, D. E. (1980). Schemata: The building blocks of cognition. In R. J. Spiro, B. Bruce & W. F. Brewer (Eds.), Theoretical issues in reading comprehension. Hillsdale, NJ: Lawrence Erlbaum Associates.■ Rumelhart, D. E., & J. L. McClelland (1986). On learning the past tenses of English verbs. In J. L. McClelland & D. E. Rumelhart (Eds.), Parallel distributed processing: Explorations in the microstructure of cognition (Vol. 2). Cambridge, MA: MIT Press.■ Rumelhart, D. E., P. Smolensky, J. L. McClelland & G. E. Hinton (1986). Schemata and sequential thought processes in PDP models. In J. L. McClelland, D. E. Rumelhart & the PDP Research Group (Eds.), Parallel Distributed Processing (Vol. 2, pp. 7-57). Cambridge, MA: MIT Press.■ Russell, B. (1927). An outline of philosophy. London: G. Allen & Unwin.■ Russell, B. (1961). History of Western philosophy. London: George Allen & Unwin.■ Russell, B. (1965). How I write. In Portraits from memory and other essays. London: Allen & Unwin.■ Russell, B. (1992). In N. Griffin (Ed.), The selected letters of Bertrand Russell (Vol. 1), The private years, 1884- 1914. Boston: Houghton Mifflin. Ryecroft, C. (1966). Psychoanalysis observed. London: Constable.■ Sagan, C. (1978). The dragons of Eden: Speculations on the evolution of human intel ligence. New York: Ballantine Books.■ Salthouse, T. A. (1992). Expertise as the circumvention of human processing limitations. In K. A. Ericsson & J. Smith (Eds.), Toward a general theory of expertise: Prospects and limits (pp. 172-194). Cambridge: Cambridge University Press.■ Sanford, A. J. (1987). The mind of man: Models of human understanding. New Haven, CT: Yale University Press.■ Sapir, E. (1921). Language. New York: Harcourt, Brace, and World.■ Sapir, E. (1964). Culture, language, and personality. Berkeley: University of California Press. (Originally published in 1941.)■ Sapir, E. (1985). The status of linguistics as a science. In D. G. Mandelbaum (Ed.), Selected writings of Edward Sapir in language, culture and personality (pp. 160166). Berkeley: University of California Press. (Originally published in 1929).■ Scardmalia, M., & C. Bereiter (1992). Literate expertise. In K. A. Ericsson & J. Smith (Eds.), Toward a general theory of expertise: Prospects and limits (pp. 172-194). Cambridge: Cambridge University Press.■ Schafer, R. (1954). Psychoanalytic interpretation in Rorschach testing. New York: Grune & Stratten.■ Schank, R. C. (1973). Identification of conceptualizations underlying natural language. In R. C. Schank & K. M. Colby (Eds.), Computer models of thought and language (pp. 187-248). San Francisco: W. H. Freeman.■ Schank, R. C. (1976). The role of memory in language processing. In C. N. Cofer (Ed.), The structure of human memory. (pp. 162-189) San Francisco: W. H. Freeman.■ Schank, R. C. (1986). Explanation patterns: Understanding mechanically and creatively. Hillsdale, NJ: Lawrence Erlbaum Associates.■ Schank, R. C., & R. P. Abelson (1977). Scripts, plans, goals, and understanding. Hillsdale, NJ: Lawrence Erlbaum Associates.■ SchroЁdinger, E. (1951). Science and humanism. Cambridge: Cambridge University Press.■ Searle, J. R. (1981a). Minds, brains, and programs. In J. Haugeland (Ed.), Mind design: Philosophy, psychology, artificial intelligence (pp. 282-306). Cambridge, MA: MIT Press.■ Searle, J. R. (1981b). Minds, brains and programs. In D. Hofstadter & D. Dennett (Eds.), The mind's I (pp. 353-373). New York: Basic Books.■ Searle, J. R. (1983). Intentionality. New York: Cambridge University Press.■ Serres, M. (1982). The origin of language: Biology, information theory, and thermodynamics. M. Anderson (Trans.). In J. V. Harari & D. F. Bell (Eds.), Hermes: Literature, science, philosophy (pp. 71-83). Baltimore: Johns Hopkins University Press.■ Simon, H. A. (1966). Scientific discovery and the psychology of problem solving. In R. G. Colodny (Ed.), Mind and cosmos: Essays in contemporary science and philosophy (pp. 22-40). Pittsburgh: University of Pittsburgh Press.■ Simon, H. A. (1979). Models of thought. New Haven, CT: Yale University Press.■ Simon, H. A. (1989). The scientist as a problem solver. In D. Klahr & K. Kotovsky (Eds.), Complex information processing: The impact of Herbert Simon. Hillsdale, N.J.: Lawrence Erlbaum Associates.■ Simon, H. A., & C. Kaplan (1989). Foundations of cognitive science. In M. Posner (Ed.), Foundations of cognitive science (pp. 1-47). Cambridge, MA: MIT Press.■ Simonton, D. K. (1988). Creativity, leadership and chance. In R. J. Sternberg (Ed.), The nature of creativity. Cambridge: Cambridge University Press.■ Skinner, B. F. (1974). About behaviorism. New York: Knopf.■ Smith, E. E. (1988). Concepts and thought. In J. Sternberg & E. E. Smith (Eds.), The psychology of human thought (pp. 19-49). Cambridge: Cambridge University Press.■ Smith, E. E. (1990). Thinking: Introduction. In D. N. Osherson & E. E. Smith (Eds.), Thinking. An invitation to cognitive science. (Vol. 3, pp. 1-2). Cambridge, MA: MIT Press.■ Socrates. (1958). Meno. In E. H. Warmington & P. O. Rouse (Eds.), Great dialogues of Plato W.H.D. Rouse (Trans.). New York: New American Library. (Original publication date unknown.)■ Solso, R. L. (1974). Theories of retrieval. In R. L. Solso (Ed.), Theories in cognitive psychology. Potomac, MD: Lawrence Erlbaum Associates.■ Spencer, H. (1896). The principles of psychology. New York: Appleton-CenturyCrofts.■ Steiner, G. (1975). After Babel: Aspects of language and translation. New York: Oxford University Press.■ Sternberg, R. J. (1977). Intelligence, information processing, and analogical reasoning. Hillsdale, NJ: Lawrence Erlbaum Associates.■ Sternberg, R. J. (1994). Intelligence. In R. J. Sternberg, Thinking and problem solving. San Diego: Academic Press.■ Sternberg, R. J., & J. E. Davidson (1985). Cognitive development in gifted and talented. In F. D. Horowitz & M. O'Brien (Eds.), The gifted and talented (pp. 103-135). Washington, DC: American Psychological Association.■ Storr, A. (1993). The dynamics of creation. New York: Ballantine Books. (Originally published in 1972.)■ Stumpf, S. E. (1994). Philosophy: History and problems (5th ed.). New York: McGraw-Hill.■ Sulloway, F. J. (1996). Born to rebel: Birth order, family dynamics, and creative lives. New York: Random House/Vintage Books.■ Thorndike, E. L. (1906). Principles of teaching. New York: A. G. Seiler.■ Thorndike, E. L. (1970). Animal intelligence: Experimental studies. Darien, CT: Hafner Publishing Co. (Originally published in 1911.)■ Titchener, E. B. (1910). A textbook of psychology. New York: Macmillan.■ Titchener, E. B. (1914). A primer of psychology. New York: Macmillan.■ Toulmin, S. (1957). The philosophy of science. London: Hutchinson.■ Tulving, E. (1972). Episodic and semantic memory. In E. Tulving & W. Donaldson (Eds.), Organisation of memory. London: Academic Press.■ Turing, A. (1946). In B. E. Carpenter & R. W. Doran (Eds.), ACE reports of 1946 and other papers. Cambridge, MA: MIT Press.■ Turkle, S. (1984). Computers and the second self: Computers and the human spirit. New York: Simon & Schuster.■ Tyler, S. A. (1978). The said and the unsaid: Mind, meaning, and culture. New York: Academic Press.■ van Heijenoort (Ed.) (1967). From Frege to Goedel. Cambridge: Harvard University Press.■ Varela, F. J. (1984). The creative circle: Sketches on the natural history of circularity. In P. Watzlawick (Ed.), The invented reality (pp. 309-324). New York: W. W. Norton.■ Voltaire (1961). On the Penseґs of M. Pascal. In Philosophical letters (pp. 119-146). E. Dilworth (Trans.). Indianapolis: Bobbs-Merrill.■ Wagman, M. (1991a). Artificial intelligence and human cognition: A theoretical inter comparison of two realms of intellect. Westport, CT: Praeger.■ Wagman, M. (1991b). Cognitive science and concepts of mind: Toward a general theory of human and artificial intelligence. Westport, CT: Praeger.■ Wagman, M. (1993). Cognitive psychology and artificial intelligence: Theory and re search in cognitive science. Westport, CT: Praeger.■ Wagman, M. (1995). The sciences of cognition: Theory and research in psychology and artificial intelligence. Westport, CT: Praeger.■ Wagman, M. (1996). Human intellect and cognitive science: Toward a general unified theory of intelligence. Westport, CT: Praeger.■ Wagman, M. (1997a). Cognitive science and the symbolic operations of human and artificial intelligence: Theory and research into the intellective processes. Westport, CT: Praeger.■ Wagman, M. (1997b). The general unified theory of intelligence: Central conceptions and specific application to domains of cognitive science. Westport, CT: Praeger.■ Wagman, M. (1998a). Cognitive science and the mind- body problem: From philosophy to psychology to artificial intelligence to imaging of the brain. Westport, CT: Praeger.■ Wagman, M. (1998b). Language and thought in humans and computers: Theory and research in psychology, artificial intelligence, and neural science. Westport, CT: Praeger.■ Wagman, M. (1998c). The ultimate objectives of artificial intelligence: Theoretical and research foundations, philosophical and psychological implications. Westport, CT: Praeger.■ Wagman, M. (1999). The human mind according to artificial intelligence: Theory, re search, and implications. Westport, CT: Praeger.■ Wagman, M. (2000). Scientific discovery processes in humans and computers: Theory and research in psychology and artificial intelligence. Westport, CT: Praeger.■ Wall, R. (1972). Introduction to mathematical linguistics. Englewood Cliffs, NJ: Prentice-Hall.■ Wallas, G. (1926). The Art of Thought. New York: Harcourt, Brace & Co.■ Wason, P. (1977). Self contradictions. In P. Johnson-Laird & P. Wason (Eds.), Thinking: Readings in cognitive science. Cambridge: Cambridge University Press.■ Wason, P. C., & P. N. Johnson-Laird. (1972). Psychology of reasoning: Structure and content. Cambridge, MA: Harvard University Press.■ Watson, J. (1930). Behaviorism. New York: W. W. Norton.■ Watzlawick, P. (1984). Epilogue. In P. Watzlawick (Ed.), The invented reality. New York: W. W. Norton, 1984.■ Weinberg, S. (1977). The first three minutes: A modern view of the origin of the uni verse. New York: Basic Books.■ Weisberg, R. W. (1986). Creativity: Genius and other myths. New York: W. H. Freeman.■ Weizenbaum, J. (1976). Computer power and human reason: From judgment to cal culation. San Francisco: W. H. Freeman.■ Wertheimer, M. (1945). Productive thinking. New York: Harper & Bros.■ Whitehead, A. N. (1925). Science and the modern world. New York: Macmillan.■ Whorf, B. L. (1956). In J. B. Carroll (Ed.), Language, thought and reality: Selected writings of Benjamin Lee Whorf. Cambridge, MA: MIT Press.■ Whyte, L. L. (1962). The unconscious before Freud. New York: Anchor Books.■ Wiener, N. (1954). The human use of human beings. Boston: Houghton Mifflin.■ Wiener, N. (1964). God & Golem, Inc.: A comment on certain points where cybernetics impinges on religion. Cambridge, MA: MIT Press.■ Winograd, T. (1972). Understanding natural language. New York: Academic Press.■ Winston, P. H. (1987). Artificial intelligence: A perspective. In E. L. Grimson & R. S. Patil (Eds.), AI in the 1980s and beyond (pp. 1-12). Cambridge, MA: MIT Press.■ Winston, P. H. (Ed.) (1975). The psychology of computer vision. New York: McGrawHill.■ Wittgenstein, L. (1953). Philosophical investigations. Oxford: Basil Blackwell.■ Wittgenstein, L. (1958). The blue and brown books. New York: Harper Colophon.■ Woods, W. A. (1975). What's in a link: Foundations for semantic networks. In D. G. Bobrow & A. Collins (Eds.), Representations and understanding: Studies in cognitive science (pp. 35-84). New York: Academic Press.■ Woodworth, R. S. (1938). Experimental psychology. New York: Holt; London: Methuen (1939).■ Wundt, W. (1904). Principles of physiological psychology (Vol. 1). E. B. Titchener (Trans.). New York: Macmillan.■ Wundt, W. (1907). Lectures on human and animal psychology. J. E. Creighton & E. B. Titchener (Trans.). New York: Macmillan.■ Young, J. Z. (1978). Programs of the brain. New York: Oxford University Press.■ Ziman, J. (1978). Reliable knowledge: An exploration of the grounds for belief in science. Cambridge: Cambridge University Press.Historical dictionary of quotations in cognitive science > Bibliography
- 1
- 2
См. также в других словарях:
Reasoning — is the cognitive process of looking for reasons for beliefs, conclusions, actions or feelings. [ Kirwin, Christopher. 1995. Reasoning . In Ted Honderich (ed.), The Oxford Companion to Philosophy . Oxford: Oxford University Press: p. 748] Humans… … Wikipedia
Formal epistemology — is a subdiscipline of epistemology that utilizes formal methods from logic, probability theory and computability theory to elucidate traditional epistemic problems. TopicsSome of the topics that come under the heading of formal epistemology… … Wikipedia
Formal equivalence checking — process is a part of electronic design automation (EDA), commonly used during the development of digital integrated circuits, to formally prove that two representations of a circuit design exhibit exactly the same behavior. Equivalence checking… … Wikipedia
formal logic — n. the branch of logic that examines patterns of reasoning to determine which ones necessarily result in valid, or formally correct, conclusions … English World dictionary
Formal interpretation — A formal interpretation [http://books.google.com/books?id=weKqT3ka5g0C pg=PA74 lpg=PA74 dq=%22Formal+interpretation%22+%22formal+language%22 source=web ots=pLN ms7Wi2 sig=P JqwdzOqLcX4nMpP64qmacnkDU hl=en#PPA74,M1 Cann Ronnie, Formal Semantics:… … Wikipedia
formal logic — the branch of logic concerned exclusively with the principles of deductive reasoning and with the form rather than the content of propositions. [1855 60] * * * Introduction the abstract study of propositions, statements, or assertively used … Universalium
Formal proof — See also: mathematical proof, proof theory, and axiomatic system A formal proof or derivation is a finite sequence of sentences (called well formed formulas in the case of a formal language) each of which is an axiom or follows from the… … Wikipedia
reasoning — noun ADJECTIVE ▪ careful, sound ▪ circular, faulty, flawed, specious (formal) ▪ underlying ▪ … Collocations dictionary
Formal fallacy — In philosophy, a formal fallacy or a logical fallacy is a pattern of reasoning which is always wrong. This is due to a flaw in the structure of the argument which renders the argument invalid. A formal fallacy is contrasted with an informal… … Wikipedia
reasoning — Any process of drawing a conclusion from a set of premises may be called a process of reasoning. If the conclusion concerns what to do, the process is called practical reasoning, otherwise pure or theoretical reasoning. Evidently such processes… … Philosophy dictionary
Formal science — A formal science is a theoretical study that is concerned with theoretical formal systems, for instance, logic, mathematics, systems theory and the theoretical branches of computer science, information theory, economics, statistics, and… … Wikipedia